AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Memory efficient

# Memory efficient

Gemma 3 1b It Fast GUFF
Quantized version optimized for low-end hardware and CPU-only environments, achieving production-ready inference configurations under resource constraints
Large Language Model
G
h4shy
101
1
Llama 3.2 3B Instruct GGUF
Llama-3.2-3B-Instruct GGUF is a 3B-parameter large language model released by Meta, utilizing IQ-DynamicGate technology for ultra-low-bit quantization (1-2 bits), optimizing inference performance while maintaining memory efficiency.
Large Language Model Supports Multiple Languages
L
Mungert
656
3
Llama 3 3 Nemotron Super 49B V1 GGUF
Other
A 49B-parameter large language model utilizing IQ-DynamicGate ultra-low-bit quantization technology, supporting 1-2 bit precision-adaptive quantization, optimized for memory efficiency and inference speed
Large Language Model English
L
Mungert
434
4
Ruleanalbert
Apache-2.0
RuLeanALBERT is a memory-efficient masked language model pretrained specifically for Russian.
Large Language Model Transformers Other
R
yandex
80
35
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase